PAGE: Robust Pattern Guided Estimation of Large Covariance Matrix
نویسندگان
چکیده
We study the problem of estimating large covariance matrices under two types of structural assumptions: (i) The covariance matrix is the summation of a low rank matrix and a sparse matrix, and we have some prior information on the sparsity pattern of the sparse matrix; (ii) The data follow a transelliptical distribution. The former structure regulates the parameter space and has its roots in different statistical models (e.g., approximate factor model, spike covariance model, and random effects model) and is motivated by some observations in financial data. The latter structure regulates the data distributions and has its root in copula modeling. Under these assumptions we propose a PAttern Guided Estimation (PAGE) method for estimating the (latent) covariance matrix. The PAGE method is rank based and naturally handles heavy tailed data. Theoretically, we show that: (i) PAGE enjoys the oracle property, i.e., it can recover the latent covariance matrix as if we know the exact low rank structure in advance; (ii) PAGE attains a fast rate of convergence as if the estimation is conducted under the Gaussian distributed data. We further extend PAGE to the situation in which the sparsity pattern is unknown. In this case, our method can be regarded as a robust version of POET in Fan et al. (2013), but is implemented through PAGE. Keyword: Matrix decomposition; Low rank matrix; Sparse matrix; Transelliptical distribution; Sparsity pattern; Catoni’s estimator; Kendall’s tau; Oracle property.
منابع مشابه
An `∞ Eigenvector Perturbation Bound and Its Application to Robust Covariance Estimation
In statistics and machine learning, people are often interested in the eigenvectors (or singular vectors) of certain matrices (e.g. covariance matrices, data matrices, etc). However, those matrices are usually perturbed by noises or statistical errors, either from random sampling or structural patterns. One usually employs Davis-Kahan sin θ theorem to bound the difference between the eigenvecto...
متن کاملRobust M-Estimation for Array Processing: A Random Matrix Approach
This article studies the limiting behavior of a robust M-estimator of population covariance matrices as both the number of available samples and the population size are large. Using tools from random matrix theory, we prove that the difference between the sample covariance matrix and (a scaled version of) the robust M-estimator tends to zero in spectral norm, almost surely. This result is appli...
متن کاملRobust Inverse Covariance Estimation under Noisy Measurements
This paper proposes a robust method to estimate the inverse covariance under noisy measurements. The method is based on the estimation of each column in the inverse covariance matrix independently via robust regression, which enables parallelization. Different from previous linear programming based methods that cannot guarantee a positive semi-definite covariance matrix, our method adjusts the ...
متن کاملPSF Estimation via Covariance Matching
This paper proposes a new covariance matching based technique for blurred image PSF (point spread function) estimation. A patch based image degradation model is proposed for the covariance matching estimation framework. A robust covariance metric which is based on Riemannian manifold is adapted to measure the distance between covariance matrices. The optimal PSF is computed by minimizing the di...
متن کاملEllipsoids for Anomaly Detection in Remote Sensing Imagery
For many target and anomaly detection algorithms, a key step is the estimation of a centroid (relatively easy) and a covariance matrix (somewhat harder) that characterize the background clutter. For a background that can be modeled as a multivariate Gaussian, the centroid and covariance lead to an explicit probability density function that can be used in likelihood ratio tests for optimal detec...
متن کامل